LLM 25-Day Course
python · 25-day course · Starts from Beginner
Day 1: Essential AI Terminology
BeginnerA comprehensive overview of key terms from AI, ML, DL differences to parameters, epochs, and loss functions
Day 2: NLP Fundamentals and Terminology
BeginnerKey NLP terms including tokens, corpus, embeddings, and attention explained with examples
Day 3: Understanding Tokenization and Embeddings
BeginnerUnderstand BPE and WordPiece algorithms and Word2Vec concepts, with hands-on practice using tiktoken
Day 4: Complete Guide to Transformer Architecture
IntermediateExplaining the Transformer's encoder/decoder structure, Self-Attention, FFN, and normalization with diagrams
Day 5: Deep Dive into Attention Mechanisms
IntermediateImplement Q/K/V, Scaled Dot-Product, Multi-Head Attention, and positional encoding from scratch with numpy
Day 6: What Are LLMs? The Principles of Large Language Models
IntermediateUnderstanding the essence of LLMs from scaling laws and pre-training to parameter-scale characteristics and emergent abilities
Day 7: OpenAI GPT Series
IntermediateModel selection guide for the latest OpenAI lineup, Responses API usage, and cost management tips
Day 8: Claude Series (Anthropic)
IntermediateExplore the Claude model lineup selection criteria, Constitutional AI concepts, and Anthropic Messages API usage
Day 9: Meta Llama Series
IntermediateUnderstand the evolution of the Llama family and run the latest open models locally
Day 10: Open-Source LLM Ecosystem Overview
IntermediateA comprehensive overview of open-source LLMs including Qwen, DeepSeek, Phi, Yi, and Korean-language models SOLAR and EXAONE
Day 11: Multimodal Models
AdvancedLearn how to process images and text together using commercial and open-source multimodal model families
Day 12: Introduction to Hugging Face Ecosystem
BeginnerA comprehensive look at the Hugging Face ecosystem including Hub, Spaces, Transformers, Datasets, and PEFT
Day 13: Getting Started with Transformers Library
BeginnerFrom installation to hands-on experience with pipeline() for text classification, sentiment analysis, translation, and summarization.
Day 14: Practical Text Generation
IntermediateUnderstand the key parameters of model.generate() and compare generation strategies including temperature, top_p, and top_k through hands-on practice.
Day 15: Advanced Tokenizer
IntermediateDeep dive into encode/decode, special tokens, padding/truncation strategies, and chat_template with apply_chat_template().
Day 16: Prompt Engineering
IntermediateLearn Zero-shot, Few-shot, Chain-of-Thought techniques, structured output (JSON), and 10 practical prompt tips.
Day 17: Building a RAG Pipeline
AdvancedBuild a complete RAG pipeline from document loading to Retrieval-Augmented Generation using LangChain and ChromaDB.
Day 18: Fine-Tuning Concepts and Strategies
IntermediateUnderstand the difference between Full Fine-Tuning and Parameter-Efficient FT, and learn decision criteria for fine-tuning vs prompting.
Day 19: Understanding LoRA and QLoRA
AdvancedLearn the principles of LoRA (Low-Rank Adaptation), rank/alpha parameters, and QLoRA's 4-bit quantization approach through code.
Day 20: PEFT Library in Practice
AdvancedHands-on practice with LoraConfig setup, target module selection, trainable parameter verification, and model saving/loading using the PEFT library.
Day 21: Preparing Fine-Tuning Datasets
IntermediateLearn Alpaca and ShareGPT dataset formats, data collection/cleaning/validation methods, and datasets library usage.
Day 22: SFT (Supervised Fine-Tuning) in Practice
AdvancedExecute actual training with trl library's SFTTrainer, and cover wandb monitoring and checkpoint management.
Day 23: Quantization Guide
AdvancedCompare precision levels from FP32 to INT4, explore GGUF/AWQ/GPTQ formats, and learn quantization methods with bitsandbytes and llama.cpp.
Day 24: Local Model Serving
IntermediateServe LLMs locally with Ollama, vLLM, Text Generation WebUI, and llama.cpp, and compare their performance.
Day 25: Mini Project -- Build Your Own AI Assistant
AdvancedIntegrate everything learned throughout the course: from model selection to fine-tuning, RAG, serving, and Gradio UI.